Skip to content
This repository was archived by the owner on Sep 10, 2025. It is now read-only.

Conversation

yanbing-j
Copy link
Contributor

@yanbing-j yanbing-j commented Dec 9, 2024

@pytorch-bot
Copy link

pytorch-bot bot commented Dec 9, 2024

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/torchchat/1404

Note: Links to docs will display an error until the docs builds have been completed.

❗ 1 Active SEVs

There are 1 currently active SEVs. If your PR is affected, please view them below:

❌ 4 New Failures

As of commit e7b6f14 with merge base bb72b09 (image):

NEW FAILURES - The following jobs have failed:

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Meta Open Source bot. label Dec 9, 2024
@Jack-Khuu
Copy link
Contributor

Jack-Khuu commented Dec 9, 2024

I'll bump the pin real quick

#1407

)

from torchao.dtypes.utils import is_device
from torchao.utils import TORCH_VERSION_AT_LEAST_2_6
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

torchchat locks onto a specific torch version, so we don't need to check

Assume > 2.6

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The CI failures seem that torchao version is not that new, because TORCH_VERSION_AT_LEAST_2_6 is a new one. And I saw you pin pytorch nightly to 20241013, which is also not new, and this nightly does not have pytorch/pytorch#139611 inside. This is my question, because the nightly used in the CI is 20241126.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

yup, working on the bump here: #1367

We'll test your fixes on there

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks!

weight = torch.empty(
(
out_features,
in_features // 2,
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

nice

out_features=out_features,
bias=False,
device="meta",
device="cpu",
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Let's keep this as a meta device as long as we can

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@Jack-Khuu Jack-Khuu added the Quantization Issues related to Quantization or torchao label Dec 10, 2024
@Jack-Khuu
Copy link
Contributor

I did a quick rebase for you; feel free to change as needed

@yanbing-j
Copy link
Contributor Author

Hi @Jack-Khuu , thanks for the rebase! The remaining 4 CI failures seem related to cuda device and I cannot get the obvious errors related to Int4 code change. Could you please help me find out the simple reproducer? Thanks!

@Jack-Khuu
Copy link
Contributor

The cuda failures are known issues

thanks for the fix

@Jack-Khuu Jack-Khuu merged commit 56be609 into pytorch:main Dec 17, 2024
49 of 53 checks passed
@yanbing-j yanbing-j deleted the yanbing/fix_1389 branch December 17, 2024 05:03
vmpuri pushed a commit that referenced this pull request Feb 4, 2025
* Update int4pack related for gguf

* Update gguf_loader.py

---------

Co-authored-by: Jack-Khuu <[email protected]>
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.

Labels

CLA Signed This label is managed by the Meta Open Source bot. Quantization Issues related to Quantization or torchao

Projects

None yet

Development

Successfully merging this pull request may close these issues.

Working around new int4wo weight packing

3 participants